Goto

Collaborating Authors

 synonym replacement


Mitigating Self-Preference by Authorship Obfuscation

Mahbub, Taslim, Feng, Shi

arXiv.org Artificial Intelligence

Language models (LMs) judges are widely used to evaluate the quality of LM outputs. Despite many advantages, LM judges display concerning biases that can impair their integrity in evaluations. One such bias is self-preference: LM judges preferring their own answers over those produced by other LMs or humans. The bias is hard to eliminate as frontier LM judges can distinguish their own outputs from those of others, even when the evaluation candidates are not labeled with their sources. In this paper, we investigate strategies to mitigate self-preference by reducing the LM judges' ability to recognize their own outputs. We apply black-box perturbations to evaluation candidates in pairwise comparison to obfuscate the authorship and reduce self-recognition. We find that perturbations as simple as synonym replacement for a few words predictably reduce self-preference. However, we also uncover fundamental challenges to eliminating the bias: when we extrapolate our perturbations to a more complete neutralization of stylistic differences between the evaluation candidates, self-preference recovers. Our findings suggest that self-recognition and self-preference can happen on many semantic levels, and complete mitigation remains challenging despite promising initial results.


Explainable AI: XAI-Guided Context-Aware Data Augmentation

Mersha, Melkamu Abay, Yigezu, Mesay Gemeda, Tonja, Atnafu Lambebo, Shakil, Hassan, Iskander, Samer, Kolesnikova, Olga, Kalita, Jugal

arXiv.org Artificial Intelligence

Explainable AI: XAI-Guided Context-A ware Data Augmentation Melkamu Abay Mersha a,, Mesay Gemeda Yigezu b, Atnafu Lambebo Tonja c, Hassan Shakil a, Samer Iskander a, Olga Kolesnikova b, Jugal Kalita a a College of Engineering and Applied Science, University of Colorado Colorado Springs, Colorado Springs, 80918, CO, USA b Instituto Polit ecnico Nacional (IPN), Centro de Investigaci on en Computaci on (CIC), 07738, Mexico City, Mexico c Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), Abu Dhabi, UAEAbstract Explainable AI (XAI) has emerged as a powerful tool for improving the performance of AI models, going beyond providing model transparency and interpretability. The scarcity of labeled data remains a fundamental challenge in developing robust and gener-alizable AI models, particularly for low-resource languages. Conventional data augmentation techniques introduce noise, cause semantic drift, disrupt contextual coherence, lack control, and lead to overfitting. To address these challenges, we propose XAI-Guided Context-A ware Data Augmentation. This novel framework leverages XAI techniques to modify less critical features while selectively preserving most task-relevant features. Our approach integrates an iterative feedback loop, which refines augmented data over multiple augmentation cycles based on explainability-driven insights and the model performance gain. Our experimental results demonstrate that XAI-SR-BT and XAI-PR-BT improve the accuracy of models on hate speech and sentiment analysis tasks by 6.6% and 8.1%, respectively, compared to the baseline, using the Amharic dataset with the XLM-R model. XAI-SR-BT and XAI-PR-BT outperform existing augmentation techniques by 4.8% and 5%, respectively, on the same dataset and model. Overall, XAI-SR-BT and XAI-PR-BT consistently outperform both baseline and conventional augmentation techniques across all tasks and models. This study provides a more controlled, interpretable, and context-aware solution to data augmentation, addressing critical limitations of existing augmentation techniques and offering a new paradigm shift for leveraging XAI techniques to enhance AI model training. Introduction The rapid advancement of large language models (LLMs), such as GPT [1] and BERT [2], has transformed various domains, including safety-critical applications. Despite their impressive capabilities, these models operate as black boxes, raising concerns about transparency, trustworthiness, and in-terpretability. Explainable Artificial Intelligence (XAI) has emerged as a key solution to these concerns, offering insights into the decision-making processes of AI models.


BDA: Bangla Text Data Augmentation Framework

Tariquzzaman, Md., Anam, Audwit Nafi, Haque, Naimul, Kabir, Mohsinul, Mahmud, Hasan, Hasan, Md Kamrul

arXiv.org Artificial Intelligence

Data augmentation involves generating synthetic samples that resemble those in a given dataset. In resource-limited fields where high-quality data is scarce, augmentation plays a crucial role in increasing the volume of training data. This paper introduces a Bangla Text Data Augmentation (BDA) Framework that uses both pre-trained models and rule-based methods to create new variants of the text. A filtering process is included to ensure that the new text keeps the same meaning as the original while also adding variety in the words used. We conduct a comprehensive evaluation of the framework's effectiveness in Bangla text classification tasks. Our framework achieved significant improvement in F1 scores across five distinct datasets, delivering performance equivalent to models trained on 100% of the data while utilizing only 50% of the training dataset. Additionally, we explore the impact of data scarcity by progressively reducing the training data and augmenting it through BDA, resulting in notable F1 score enhancements. The study offers a thorough examination of BDA's performance, identifying key factors for optimal results and addressing its limitations through detailed analysis.


Papilusion at DAGPap24: Paper or Illusion? Detecting AI-generated Scientific Papers

Andreev, Nikita, Shirnin, Alexander, Mikhailov, Vladislav, Artemova, Ekaterina

arXiv.org Artificial Intelligence

This paper presents Papilusion, an AI-generated scientific text detector developed within the DAGPap24 shared task on detecting automatically generated scientific papers. We propose an ensemble-based approach and conduct ablation studies to analyze the effect of the detector configurations on the performance. Papilusion is ranked 6th on the leaderboard, and we improve our performance after the competition ended, achieving 99.46 (+9.63) of the F1-score on the official test set.


MPAT: Building Robust Deep Neural Networks against Textual Adversarial Attacks

Zhang, Fangyuan, Zhou, Huichi, Li, Shuangjiao, Wang, Hongtao

arXiv.org Artificial Intelligence

Deep neural networks have been proven to be vulnerable to adversarial examples and various methods have been proposed to defend against adversarial attacks for natural language processing tasks. However, previous defense methods have limitations in maintaining effective defense while ensuring the performance of the original task. In this paper, we propose a malicious perturbation based adversarial training method (MPAT) for building robust deep neural networks against textual adversarial attacks. Specifically, we construct a multi-level malicious example generation strategy to generate adversarial examples with malicious perturbations, which are used instead of original inputs for model training. Additionally, we employ a novel training objective function to ensure achieving the defense goal without compromising the performance on the original task. We conduct comprehensive experiments to evaluate our defense method by attacking five victim models on three benchmark datasets. The result demonstrates that our method is more effective against malicious adversarial attacks compared with previous defense methods while maintaining or further improving the performance on the original task.


Distributional Data Augmentation Methods for Low Resource Language

Mahamud, Mosleh, Lee, Zed, Samsten, Isak

arXiv.org Artificial Intelligence

Text augmentation is a technique for constructing synthetic data from an under-resourced corpus to improve predictive performance. Synthetic data generation is common in numerous domains. However, recently text augmentation has emerged in natural language processing (NLP) to improve downstream tasks. One of the current state-of-the-art text augmentation techniques is easy data augmentation (EDA), which augments the training data by injecting and replacing synonyms and randomly permuting sentences. One major obstacle with EDA is the need for versatile and complete synonym dictionaries, which cannot be easily found in low-resource languages. To improve the utility of EDA, we propose two extensions, easy distributional data augmentation (EDDA) and type specific similar word replacement (TSSR), which uses semantic word context information and part-of-speech tags for word replacement and augmentation. In an extensive empirical evaluation, we show the utility of the proposed methods, measured by F1 score, on two representative datasets in Swedish as an example of a low-resource language. With the proposed methods, we show that augmented data improve classification performances in low-resource settings.


Data Augmentation for Low-Resource Keyphrase Generation

Garg, Krishna, Chowdhury, Jishnu Ray, Caragea, Cornelia

arXiv.org Artificial Intelligence

Keyphrase generation is the task of summarizing the contents of any given article into a few salient phrases (or keyphrases). Existing works for the task mostly rely on large-scale annotated datasets, which are not easy to acquire. Very few works address the problem of keyphrase generation in low-resource settings, but they still rely on a lot of additional unlabeled data for pretraining and on automatic methods for pseudo-annotations. In this paper, we present data augmentation strategies specifically to address keyphrase generation in purely resource-constrained domains. We design techniques that use the full text of the articles to improve both present and absent keyphrase generation. We test our approach comprehensively on three datasets and show that the data augmentation strategies consistently improve the state-of-the-art performance. We release our source code at https://github.com/kgarg8/kpgen-lowres-data-aug.


Which anonymization technique is best for which NLP task? -- It depends. A Systematic Study on Clinical Text Processing

Larbi, Iyadh Ben Cheikh, Burchardt, Aljoscha, Roller, Roland

arXiv.org Artificial Intelligence

Clinical text processing has gained more and more attention in recent years. The access to sensitive patient data, on the other hand, is still a big challenge, as text cannot be shared without legal hurdles and without removing personal information. There are many techniques to modify or remove patient related information, each with different strengths. This paper investigates the influence of different anonymization techniques on the performance of ML models using multiple datasets corresponding to five different NLP tasks. Several learnings and recommendations are presented. This work confirms that particularly stronger anonymization techniques lead to a significant drop of performance. In addition to that, most of the presented techniques are not secure against a re-identification attack based on similarity search.


Augment Your Small Dataset Using Transformers and Synonym Replacement for Sentiment Analysis-- Part…

#artificialintelligence

Its uniqueness lies in its'self-supervised', pre-training objective architecture. Unlike other models that infer on the meaning of a sentence by extracting small parts of it, Pegasus completely'masks' the sentence and tries to find it by reading the text before and after it. Pegasus is really good at data summarization, but it is also great at paraphrasing sentences. The model is extremely easy to use, doesn't require many dependencies and with just a few lines of code we'll have our augmented dataset ready for training. To be able to leverage our small dataset efficiently, we will be performing text Paraphrasing along with Synonym Replacement to come up with a dataset large and unique enough to train our Sentiment Analysis model with.